1Hiphen SAS, 120 Rue Jean Dausset, Agroparc, Bâtiment Technicité, 84140 Avignon, France
2INRAE, UMR EMMAH, UMT CAPTE, 228 Route de l'Aérodrome, Domaine Saint Paul-Site Agroparc CS 40509, 84914 Avignon Cedex 9, France
3Arvalis, 228, Route de l'Aérodrome-CS 40509, 84914 Avignon Cedex 9, France
4International Field Phenomics Research Laboratory, Institute for Sustainable Agro-ecosystem Services, Graduate School of Agricultural and Life Sciences, The University of Tokyo, Tokyo, Japan
Received 16 Apr 2021 |
Accepted 02 Jul 2021 |
Published 21 Aug 2021 |
Early-stage plant density is an essential trait that determines the fate of a genotype under given environmental conditions and management practices. The use of RGB images taken from UAVs may replace the traditional visual counting in fields with improved throughput, accuracy, and access to plant localization. However, high-resolution images are required to detect the small plants present at the early stages. This study explores the impact of image ground sampling distance (GSD) on the performances of maize plant detection at three-to-five leaves stage using Faster-RCNN object detection algorithm. Data collected at high resolution () over six contrasted sites were used for model training. Two additional sites with images acquired both at high and low () resolutions were used to evaluate the model performances. Results show that Faster-RCNN achieved very good plant detection and counting () performances when native high-resolution images are used both for training and validation. Similarly, good performances were observed () when the model is trained over synthetic low-resolution images obtained by downsampling the native training high-resolution images and applied to the synthetic low-resolution validation images. Conversely, poor performances are obtained when the model is trained on a given spatial resolution and applied to another spatial resolution. Training on a mix of high- and low-resolution images allows to get very good performances on the native high-resolution () and synthetic low-resolution () images. However, very low performances are still observed over the native low-resolution images (), mainly due to the poor quality of the native low-resolution images. Finally, an advanced super resolution method based on GAN (generative adversarial network) that introduces additional textural information derived from the native high-resolution images was applied to the native low-resolution validation images. Results show some significant improvement () compared to bicubic upsampling approach, while still far below the performances achieved over the native high-resolution images.